Algorithmic Information Theory and Computational Complexity

نویسنده

  • Rusins Freivalds
چکیده

We present examples where theorems on complexity of computation are proved using methods in algorithmic information theory. The first example is a non-effective construction of a language for which the size of any deterministic finite automaton exceeds the size of a probabilistic finite automaton with a bounded error exponentially. The second example refers to frequency computation. Frequency computation was introduced by Rose and McNaughton in early sixties and developed by Trakhtenbrot, Kinber, Degtev, Wechsung, Hinrichs and others. A transducer is a finite-state automaton with an input and an output. We consider the possibilities of probabilistic and frequency transducers and prove several theorems establishing an infinite hierarchy of relations. We consider only relations where for each input value there is exactly one allowed output value. Relations computable by weak finite-state transducers with frequency km kn but not with frequency m n are presented in a non-constructive way using methods of algorithmic information theory.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

3D Scene and Object Classification Based on Information Complexity of Depth Data

In this paper the problem of 3D scene and object classification from depth data is addressed. In contrast to high-dimensional feature-based representation, the depth data is described in a low dimensional space. In order to remedy the curse of dimensionality problem, the depth data is described by a sparse model over a learned dictionary. Exploiting the algorithmic information theory, a new def...

متن کامل

Probabilistic Sufficiency and Algorithmic Sufficiency from the point of view of Information Theory

‎Given the importance of Markov chains in information theory‎, ‎the definition of conditional probability for these random processes can also be defined in terms of mutual information‎. ‎In this paper‎, ‎the relationship between the concept of sufficiency and Markov chains from the perspective of information theory and the relationship between probabilistic sufficiency and algorithmic sufficien...

متن کامل

Algorithmic Information Theory and Kolmogorov Complexity

This document contains lecture notes of an introductory course on Kolmogorov complexity. They cover basic notions of algorithmic information theory: Kolmogorov complexity (plain, conditional, prefix), notion of randomness (Martin-Löf randomness, Mises–Church randomness), Solomonoff universal a priori probability and their properties (symmetry of information, connection between a priori probabil...

متن کامل

Multisource Algorithmic Information Theory

Multisource information theory in Shannon setting is well known. In this article we try to develop its algorithmic information theory counterpart and use it as the general framework for many interesting questions about Kolmogorov complexity.

متن کامل

Two-dimensional Kolmogorov complexity and an empirical validation of the Coding theorem method by compressibility

We propose a measure based upon the fundamental theoretical concept in algorithmic information theory that provides a natural approach to the problem of evaluating n-dimensional complexity by using an n-dimensional deterministic Turing machine. The technique is interesting because it provides a natural algorithmic process for symmetry breaking generating complex n-dimensional structures from pe...

متن کامل

Recurrence and algorithmic information

In this paper we initiate a somewhat detailed investigation of the relationships between quantitative recurrence indicators and algorithmic complexity of orbits in weakly chaotic dynamical systems. We mainly focus on examples.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2011